Generalization of value in reinforcement learning by humans.
نویسندگان
چکیده
Research in decision-making has focused on the role of dopamine and its striatal targets in guiding choices via learned stimulus-reward or stimulus-response associations, behavior that is well described by reinforcement learning theories. However, basic reinforcement learning is relatively limited in scope and does not explain how learning about stimulus regularities or relations may guide decision-making. A candidate mechanism for this type of learning comes from the domain of memory, which has highlighted a role for the hippocampus in learning of stimulus-stimulus relations, typically dissociated from the role of the striatum in stimulus-response learning. Here, we used functional magnetic resonance imaging and computational model-based analyses to examine the joint contributions of these mechanisms to reinforcement learning. Humans performed a reinforcement learning task with added relational structure, modeled after tasks used to isolate hippocampal contributions to memory. On each trial participants chose one of four options, but the reward probabilities for pairs of options were correlated across trials. This (uninstructed) relationship between pairs of options potentially enabled an observer to learn about option values based on experience with the other options and to generalize across them. We observed blood oxygen level-dependent (BOLD) activity related to learning in the striatum and also in the hippocampus. By comparing a basic reinforcement learning model to one augmented to allow feedback to generalize between correlated options, we tested whether choice behavior and BOLD activity were influenced by the opportunity to generalize across correlated options. Although such generalization goes beyond standard computational accounts of reinforcement learning and striatal BOLD, both choices and striatal BOLD activity were better explained by the augmented model. Consistent with the hypothesized role for the hippocampus in this generalization, functional connectivity between the ventral striatum and hippocampus was modulated, across participants, by the ability of the augmented model to capture participants' choice. Our results thus point toward an interactive model in which striatal reinforcement learning systems may employ relational representations typically associated with the hippocampus.
منابع مشابه
Unspecific Reinforcement Learning in One and Two - layered Networks
The dynamics of on-line learning of a perceptron with a learning rule based on the Hebb rule with “delayed” unspecific reinforcement is studied for a special case of the feedback definition. This learning algorithm combines an associative and a reinforcement step and the relevant learning parameter λ represents the ratio of the associative to the reinforcement step. For given initial conditions...
متن کاملPerception-Based Generalization in Model-Based Reinforcement Learning
OF THE DISSERTATION Perception-Based Generalization in Model-Based Reinforcement Learning by Bethany R. Leffler Dissertation Director: Michael L. Littman In recent years, the advances in robotics have allowed for robots to venture into places too dangerous for humans. Unfortunately, the terrain in which these robots are being deployed may not be known by humans in advance, making it difficult t...
متن کاملHow Perceptual Categories Influence Trial and Error Learning in Humans
Converging evidence suggests that trial-and-error learning in humans shares many computational principals with contemporary RL algorithms (Montague, Hyman, & Cohen, 2004; Frank, Seeberger, & O’Reilly, 2004). However, a critical feature of these algorithms is the notion of “state.” State representations help structure which actions to take in particular situations, as well as how to assign credi...
متن کاملNovelty and Inductive Generalization in Human Reinforcement Learning
In reinforcement learning (RL), a decision maker searching for the most rewarding option is often faced with the question: What is the value of an option that has never been tried before? One way to frame this question is as an inductive problem: How can I generalize my previous experience with one set of options to a novel option? We show how hierarchical Bayesian inference can be used to solv...
متن کاملEfficient Reinforcement Learning for Motor Control
Artificial learners often require many more trials than humans or animals when learning motor control tasks in the absence of expert knowledge. We implement two key ingredients of biological learning systems, generalization and incorporation of uncertainty into the decision-making process, to speed up artificial learning. We present a coherent and fully Bayesian framework that allows for effici...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- The European journal of neuroscience
دوره 35 7 شماره
صفحات -
تاریخ انتشار 2012